Information-theoretics Based Error-metrics for Gradient Descent Learning in Neutral Networks

نویسندگان

  • Joseph C. Park
  • Perambur S. Neelakanta
  • Salahalddin T. Abusalah
  • Dolores F. De Groff
  • Raghavan Sudhakar
چکیده

Conventionally, square error (SE) and/or relative ent ropy (RE) error functions defined over a training set are adop ted towa rds optimization of gradient descent learnings in neur al networks. As an alternative, a set of divergence (or dist ance) measur es can be specified in the inform ationtheoretic plane that funct iona lly have pragmatic values similar to (or improved upon) The SE or RE metrics. KullbackLeibler (KL) , Jensen (J ), and Jensen-Sh ann on (JS) meas ures are suggested as possible inform ation-theoretic error-met ric candidates that are defined and derived explicit ly. Both convent ional SE / RE measures , as well as the prop osed inform ation-theoreti c error-met rics, are applied to t rain a multilayer perceptron topology. This is done in order to elucida te t heir relative efficacy in deciding the perform ance of the network as evidenced from the convergence rat es and tr aining t imes involved. Pertinent simulation results are present ed and discussed.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Handwritten Character Recognition using Modified Gradient Descent Technique of Neural Networks and Representation of Conjugate Descent for Training Patterns

The purpose of this study is to analyze the performance of Back propagation algorithm with changing training patterns and the second momentum term in feed forward neural networks. This analysis is conducted on 250 different words of three small letters from the English alphabet. These words are presented to two vertical segmentation programs which are designed in MATLAB and based on portions (1...

متن کامل

Geoid Determination Based on Log Sigmoid Function of Artificial Neural Networks: (A case Study: Iran)

A Back Propagation Artificial Neural Network (BPANN) is a well-known learning algorithmpredicated on a gradient descent method that minimizes the square error involving the networkoutput and the goal of output values. In this study, 261 GPS/Leveling and 8869 gravity intensityvalues of Iran were selected, then the geoid with three methods “ellipsoidal stokes integral”,“BPANN”, and “collocation” ...

متن کامل

Designing stable neural identifier based on Lyapunov method

The stability of learning rate in neural network identifiers and controllers is one of the challenging issues which attracts great interest from researchers of neural networks. This paper suggests adaptive gradient descent algorithm with stable learning laws for modified dynamic neural network (MDNN) and studies the stability of this algorithm. Also, stable learning algorithm for parameters of ...

متن کامل

Novel Radial Basis Function Neural Networks based on Probabilistic Evolutionary and Gaussian Mixture Model for Satellites Optimum Selection

In this study, two novel learning algorithms have been applied on Radial Basis Function Neural Network (RBFNN) to approximate the functions with high non-linear order. The Probabilistic Evolutionary (PE) and Gaussian Mixture Model (GMM) techniques are proposed to significantly minimize the error functions. The main idea is concerning the various strategies to optimize the procedure of Gradient ...

متن کامل

Direct adaptive output tracking control using multilayered neural networks - Control Theory and Applications [see also IEE Proceedings-Control Theory and Applications], IEE Pr

Multilayered neural networks are used to construct nonlinear learning control systems for a class of unknown nonlinear systems in a canonical form. An adaptive output tracking architecture is proposed using the outputs of the two three-layered neutral networks which are trained to approximate the unknown nonlinear plant to any desired degree of accuracy by using the modified back-propagation te...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • Complex Systems

دوره 9  شماره 

صفحات  -

تاریخ انتشار 1995